Convergence results for the EM approach to mixtures of experts architectures

نویسندگان

  • Michael I. Jordan
  • Lei Xu
چکیده

The Expectation-Maximization (EM) algorithm is an iterative approach to maximum likelihood parameter estimation. Jordan and Jacobs (1994) recently proposed an EM algorithm for the mixture of experts architecture of Jacobs, Jordan, Nowlan and Hinton (1991) and the hierarchical mixture of experts architecture of Jordan and Jacobs (1992). They showed empirically that the EM algorithm for these architectures yields signiicantly faster convergence than gradient ascent. In the current paper we provide a theoretical analysis of this algorithm. We show that the algorithm can be regarded as a variable metric algorithm with its searching direction having a positive projection on the gradient of the log likelihood. We also analyze the convergence of the algorithm and provide an explicit expression for the convergence rate. In addition, we describe an acceleration technique that yields a signiicant speedup in simulation experiments.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An Alternative Model for Mixtures of Experts

An alternative model is proposed for mixtures of experts by utiliz ing a di erent parametric form for the gating network The mod i ed model is trained by an EM algorithm In comparison with earlier models trained by either EM or gradient ascent there is no need to select a learning stepsize to guarantee the convergence of the learning procedure We report simulation experiments which show that th...

متن کامل

Analysis of Strategies for Integrating Humanities with Medical Sciences in Iran from the Perspective of Humanities Experts with Convergence Sciences Approach

Introduction: Medical humanities are one of the most important interdisciplinary needs. Since the most important discussion in this field is how to achieve it, the purpose of this study is to analyze it from the perspective of humanities experts in the Iran. Method: The present study is a national wide qualitative study that was conducted in 2019. Considering the saturation criterion, data wer...

متن کامل

Mixture of experts architectures for neural networks as a special case of conditional expectation formula

Recently a new interesting architecture of neural networks called “mixture of experts” has been proposed as a tool of real multivariate approximation or classification. It is shown that, in some cases, the underlying problem of prediction can be solved by estimating the joint probability density of involved variables. Assuming the model of Gaussian mixtures we can explictly write the optimal mi...

متن کامل

Spectral Experts for Estimating Mixtures of Linear Regressions

Discriminative latent-variable models are typically learned using EM or gradient-based optimization, which suffer from local optima. In this paper, we develop a new computationally efficient and provably consistent estimator for a mixture of linear regressions, a simple instance of a discriminative latentvariable model. Our approach relies on a lowrank linear regression to recover a symmetric t...

متن کامل

Convergence of the EM Algorithm for Gaussian Mixtures with Unbalanced Mixing Coefficients

The speed of convergence of the Expectation Maximization (EM) algorithm for Gaussian mixture model fitting is known to be dependent on the amount of overlap among the mixture components. In this paper, we study the impact of mixing coefficients on the convergence of EM. We show that when the mixture components exhibit some overlap, the convergence of EM becomes slower as the dynamic range among...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Neural Networks

دوره 8  شماره 

صفحات  -

تاریخ انتشار 1995